Analogical Natural Language Processing
نویسندگان
چکیده
The use of examples as the basis for machine translation systems has gained considerable acceptance since the original proposal of Nagao in 1984. In this short book, Jones first reviews the fundamental principles of example-based machine translation (EBMT) in order to then introduce the specific mechanisms of his model. The key characteristic of this model is its purely stochastic processing, which is based on the algorithm put forth by Skousen (1989) in his Analogical Modeling strategy for language comprehension. Overall, the book is well-written and offers a good introduction to some of the very interesting problems of machine translation. But my reading left me somewhat unsatisfied. In order to explain this, let me first summarize the chapters. The goals and method of the work are clearly stated in the nine pages that form the introduction: Jones intends to demonstrate the possibility of uniformly using examples rather than rules for machine translation. I emphasize that, in this context, "the term 'analogy' refers to the general process of deriving information about some new piece of language by comparison with known piece(s) of language" (p. 4). Also, it is important to understand that "the rejection of rule-based natural language processing does not mean that linguistic representations cannot be used" (p. 5). Indeed, Jones subscribes to a non-inferential heavily-predicated representational strategy based on Functional Grammar. Within this framework, "source language predications and their target language equivalents are stored as instantiated examples of translation, the source half of which attempt to clone themselves onto the source language input . . . . The translations are achieved by generating the target language half of the best matching example into a surface realization . . . . Where only partial cloning is being achieved across a set of examples, the process of recombination can be used to recombine elements of example predications based on the results of the analogical modelling" (p. 6). The second chapter provides some additional background material on EBMT, as well as a partial survey of relevant research. Depth is typically privileged over breadth. For example, Functional Grammar seems to reduce exclusively to the work of Dik (1978) and van der Korst (1989), with the consequence that some details are mentioned but not explained, or discussed at length and then never revisited in the rest of the book. Chapter 3 then overviews the proposed model, and more specifically, the representations it uses. Very quickly the reader discovers complex hand-coded frames littered with the primitives of Functional Grammar, as well as somewhat artificial examples from a narrow domain. Thus, the usual criticism against complex innate (nay, ad hoc) data structures soon comes to mind. Jones does motivate the use of Functional Grammar for machine translation and raises lots of interesting questions for MT. But
منابع مشابه
Integrating Natural Language, Knowledge Representation and Reasoning, and Analogical Processing to Learn by Reading
•radically change the economics of building large knowledge bases •provide a platform for cognitive simulations of larger-scale phenomena •Learning Reader learns by reading simplified language texts •Manages syntactic complexity •Unconstrained vocabulary, unlike controlled languages •Learning Reader combines •Natural language processing •A large-scale knowledge base •Deductive reasoning •Analog...
متن کاملAn Analogical Learner for Morphological Analysis
Analogical learning is based on a twostep inference process: (i) computation of a structural mapping between a new and a memorized situation; (ii) transfer of knowledge from the known to the unknown situation. This approach requires the ability to search for and exploit such mappings, hence the need to properly define analogical relationships, and to efficiently implement their computation. In ...
متن کاملAnalogical Learning and Formal Proportions: Definitions and Methodological Issues
Analogical learning is a two-step inference process: (i) computation of a mapping between a new and a memorized situation; (ii) transfer from the known to the unknown situation. This approach requires the ability to search for and exploit such mappings, which are based on the notion of analogical proportions, hence the need to properly define these proportions, and to efficiently implement thei...
متن کاملSkousen s Analogical Modeling Algorithm A comparison with Lazy Learning
We provide a qualitative and empirical comparison of Skousen s Analogical Modeling algorithm AM with Lazy Learning LL on a typical Natural Language Processing task AM incorporates an original approach to feature selection and to the handling of sym bolic unordered feature values More speci cally it provides a method to dynamically compute an optimally sized set of nearest neighbours the analogi...
متن کاملAnalogical Word Sense Disambiguation
Word sense disambiguation is an important problem in learning by reading. This paper introduces analogical word-sense disambiguation, which uses human-like analogical processing over structured, relational representations to perform word sense disambiguation. Cases are automatically constructed using representations produced via natural language analysis of sentences, and include both conceptua...
متن کاملIssues in Analogical Inference Over Sequences of Symbols: A Case Study on Proper Name Transliteration
Formal analogies, that is, proportional analogies involving relations at a formal level (e.g. cordially is to cordial as appreciatively is to appreciative) have a long history in Linguistics. They can accommodate a wide variety of linguistic data without resorting to ad hoc representations and are inherently good at capturing long range dependencies between data. Unfortunately, applying analogi...
متن کامل